Author:sana
Released:March 16, 2026
The debate around AI often starts with the wrong question: which model type will win, open or closed? That framing doesn’t match what’s actually happening. Companies are not choosing one side. They are building around both.
Closed models act like premium software services. You pay for ease of use, stability, built-in safety controls, and support agreements. Teams can plug in an API and ship quickly without managing infrastructure.
Open-weight models play a different role. They are closer to raw infrastructure. Teams can fine-tune them, run them on their own hardware, and control how they behave and scale.
By now, usage data shows a clear split. Open-weight models handle about one-third of total token usage, while proprietary models still process the majority. This is not a transition from one to the other. It is a two-track system already in place.
Enterprise spending on generative AI is already in the multi-billion-dollar range. A large share still goes to a handful of proprietary providers through API usage and cloud contracts. These options are easy to adopt and fit existing procurement processes.
At the same time, open-weight models are gaining ground, especially through routing platforms. Many teams no longer rely on a single model. They route requests across several models depending on task type, response speed, and price.
For example, a coding assistant might call a high-end proprietary model for complex reasoning, but switch to a cheaper open model for autocomplete or formatting tasks.
Global supply is expanding quickly. Chinese labs and startups have released competitive open-weight models, increasing availability and pushing prices down. For buyers, this mainly shows that the model layer is becoming more competitive and less concentrated.
Another change is the lack of a clear leader among open models. Different models perform better on different tasks. Developers test several options and swap them frequently. Model choice is becoming a runtime decision rather than a fixed bet.
This makes the market look more like cloud services, where workloads move across providers based on cost and performance.

A year ago, closed models had a clear lead in most benchmarks.
That lead has narrowed sharply. The Stanford AI Index 2025 shows the gap shrinking from about 8% to around 1.7% on key evaluations.
For many real-world tasks, that difference is hard to notice. Open-weight models now reach over 90% of top-tier performance while costing much less per token.
This is especially visible in:
Mid-sized models are also improving fast. Models in the tens of billions of parameters now handle most enterprise workloads reliably. They run faster, require less compute, and are easier to deploy in private environments.
Closed models still lead in areas like advanced reasoning, multimodal tasks, and edge cases that require deeper context handling. For most business use, that edge is no longer the main factor in model choice.
Open models are often described as “free,” which is misleading.
The cost shifts to different areas. Instead of paying per API call, companies pay for:
There are also efficiency differences. Some studies show open-weight models can use 1.5 to 4 times more tokens for the same task. For simple queries, the gap can be even larger. More tokens mean a higher compute cost.
Closed APIs bundle everything into one price. For a team without infrastructure experience, this can be cheaper overall. There is no need to manage servers or optimize inference.
On a larger scale, the math changes. A company running millions of requests per day can reduce unit costs by hosting its own models, especially if it secures favorable hardware pricing.
A simple example:
The right choice depends on workload size, token usage patterns, and team capability.
Large companies are already combining both approaches in production.
A typical setup includes several closed APIs plus at least one open-weight model running in a private or virtual private cloud environment.
The workflow usually follows a clear pattern:
Teams begin with a top closed model to test ideas quickly. This reduces setup time and helps validate whether a use case is worth pursuing.
Once a workflow becomes stable and usage grows, teams look for ways to lower cost. High-volume tasks, such as document processing or customer support replies, are often moved to open-weight models.
Sensitive data adds another layer. Financial, healthcare, or internal corporate data is often processed using self-hosted models to meet compliance and privacy requirements.
This leads to a layered system where different models handle different parts of the pipeline.
One important takeaway from this shift is where value actually sits. It is not in the base model alone.
Value comes from:
The earlier “dumb app” concern pointed this out. Products that only wrap a model without adding data or workflow logic are easy to replace. Strong applications integrate deeply and improve over time.
Model providers supply capability, but applications capture lasting value.

There is no single winner between open and closed models.
Closed models work well for fast rollout, ease of use, and compliance-heavy environments. Open models fit teams that need control, customization, and better unit economics at scale.
AI is starting to resemble cloud infrastructure. Training may stay concentrated, but usage is spreading across many providers and setups.
The strategy is simple: stay flexible. Build systems that can switch models, match tools to each workload, and avoid locking into one provider.
The advantage will go to teams that can use many models well, not those that pick just one.